AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
4-bit quantization efficient inference

# 4-bit quantization efficient inference

Saanvi C0 12B
Apache-2.0
A 12-billion-parameter large language model optimized for speed, efficiency, and contextual accuracy, supporting RAG-enhanced technology and a 128k context window.
Large Language Model Transformers
S
riple-saanvi-lab
170
2
Meta Llama 3 8B Instruct Hf AWQ
Other
Meta Llama 3 series large language model, featuring an 8 billion parameter instruction fine-tuned text generation model optimized for dialogue scenarios.
Large Language Model Transformers
M
solidrust
848
9
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase